Simple Bayesian Classifiers Do Not Assume Independence
نویسندگان
چکیده
Bayes’ theorem tells us how to optimally predict the class of a previously unseen example, given a training sample. The chosen class should be the one which maximizes P(CilE) = P(Ci) P(EICi) /P(E), where Ci is the ith class, E is the test example, P(YIX) denotes the conditional probability of Y given X, and probabilities are estimated from the training sample. Let an example be a vector of a attributes. If the attributes are independent given the class, P(EICi) can be decomposed into the product P(vlICi) . . . P(va(Ci), where v.j is the value of the jth attribute in the example E. Therefore we should predict the class that maximizes:
منابع مشابه
A Theoretical and Experimental Evaluation of Augmented Bayesian Classifiers
Naive Bayes is a simple Bayesian network classifier with strong independence assumptions among features. This classifier despite its strong independence assumptions, often performs well in practice. It is believed that relaxing the independence assumptions of naive Bayes may improve the performance of the resulting structure. Augmented Bayesian Classifiers relax the independence assumptions of ...
متن کاملBeyond Independence: Conditions for the Optimality of the Simple Bayesian Classi er
The simple Bayesian classi er (SBC) is commonly thought to assume that attributes are independent given the class, but this is apparently contradicted by the surprisingly good performance it exhibits in many domains that contain clear attribute dependences. No explanation for this has been proposed so far. In this paper we show that the SBC does not in fact assume attribute independence, and ca...
متن کاملBeyond Independence: Conditions for the Optimality of the Simple Bayesian Classifier
The simple Bayesian classi er (SBC) is commonly thought to assume that attributes are independent given the class, but this is apparently contradicted by the surprisingly good performance it exhibits in many domains that contain clear attribute dependences. No explanation for this has been proposed so far. In this paper we show that the SBC does not in fact assume attribute independence, and ca...
متن کاملTractable Bayesian Learning of Tree Augmented Naive Bayes Classifiers
Bayesian classifiers such as Naive Bayes or Tree Augmented Naive Bayes (TAN) have shown excellent performance given their simplicity and heavy underlying independence assumptions. In this paper we introduce a classifier taking as basis the TAN models and taking into account uncertainty in model selection. To do this we introduce decomposable distributions over TANs and show that the expression ...
متن کاملSisterhood of Classifiers: A Comparative Study of Naive Bayes and Noisy-or Networks
Classification is a task central to many machine learning problems. In this paper we examine two Bayesian network classifiers, the naive Bayes and the noisy-or models. They are of particular interest because of their simple structures. We compare them on two dimensions: expressive power and ability to learn. As it turns out, naive Bayes, noisy-or, and logistic regression classifiers all have eq...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 1996